280 research outputs found
Feature Tracking Cardiac Magnetic Resonance via Deep Learning and Spline Optimization
Feature tracking Cardiac Magnetic Resonance (CMR) has recently emerged as an
area of interest for quantification of regional cardiac function from balanced,
steady state free precession (SSFP) cine sequences. However, currently
available techniques lack full automation, limiting reproducibility. We propose
a fully automated technique whereby a CMR image sequence is first segmented
with a deep, fully convolutional neural network (CNN) architecture, and
quadratic basis splines are fitted simultaneously across all cardiac frames
using least squares optimization. Experiments are performed using data from 42
patients with hypertrophic cardiomyopathy (HCM) and 21 healthy control
subjects. In terms of segmentation, we compared state-of-the-art CNN
frameworks, U-Net and dilated convolution architectures, with and without
temporal context, using cross validation with three folds. Performance relative
to expert manual segmentation was similar across all networks: pixel accuracy
was ~97%, intersection-over-union (IoU) across all classes was ~87%, and IoU
across foreground classes only was ~85%. Endocardial left ventricular
circumferential strain calculated from the proposed pipeline was significantly
different in control and disease subjects (-25.3% vs -29.1%, p = 0.006), in
agreement with the current clinical literature.Comment: Accepted to Functional Imaging and Modeling of the Heart (FIMH) 201
UPI-Net: Semantic Contour Detection in Placental Ultrasound
Semantic contour detection is a challenging problem that is often met in
medical imaging, of which placental image analysis is a particular example. In
this paper, we investigate utero-placental interface (UPI) detection in 2D
placental ultrasound images by formulating it as a semantic contour detection
problem. As opposed to natural images, placental ultrasound images contain
specific anatomical structures thus have unique geometry. We argue it would be
beneficial for UPI detectors to incorporate global context modelling in order
to reduce unwanted false positive UPI predictions. Our approach, namely
UPI-Net, aims to capture long-range dependencies in placenta geometry through
lightweight global context modelling and effective multi-scale feature
aggregation. We perform a subject-level 10-fold nested cross-validation on a
placental ultrasound database (4,871 images with labelled UPI from 49 scans).
Experimental results demonstrate that, without introducing considerable
computational overhead, UPI-Net yields the highest performance in terms of
standard contour detection metrics, compared to other competitive benchmarks.Comment: 9 pages, 8 figures, accepted at Visual Recognition for Medical Images
(VRMI), ICCV 201
Rethinking Semi-Supervised Federated Learning: How to co-train fully-labeled and fully-unlabeled client imaging data
The most challenging, yet practical, setting of semi-supervised federated
learning (SSFL) is where a few clients have fully labeled data whereas the
other clients have fully unlabeled data. This is particularly common in
healthcare settings where collaborating partners (typically hospitals) may have
images but not annotations. The bottleneck in this setting is the joint
training of labeled and unlabeled clients as the objective function for each
client varies based on the availability of labels. This paper investigates an
alternative way for effective training with labeled and unlabeled clients in a
federated setting. We propose a novel learning scheme specifically designed for
SSFL which we call Isolated Federated Learning (IsoFed) that circumvents the
problem by avoiding simple averaging of supervised and semi-supervised models
together. In particular, our training approach consists of two parts - (a)
isolated aggregation of labeled and unlabeled client models, and (b) local
self-supervised pretraining of isolated global models in all clients. We
evaluate our model performance on medical image datasets of four different
modalities publicly available within the biomedical image classification
benchmark MedMNIST. We further vary the proportion of labeled clients and the
degree of heterogeneity to demonstrate the effectiveness of the proposed method
under varied experimental settings.Comment: Published in MICCAI 2023 with early acceptance and selected as 1 of
the top 20 poster highlights under the category: Which work has the potential
to impact other applications of AI and C
Effects of Surface Morphology on the Anchoring and Electrooptical Dynamics of Confined Nanoscale Liquid Crystalline Films
The orientation and dynamics of two 40-nm thick films of 4-n-pentyl-4‘-cyanobiphenyl (5CB), a nematic liquid crystal, have been studied using step-scan Fourier transform infrared spectroscopy (FTIR). The films are confined in nanocavities bounded by an interdigitated electrode array (IDA) patterned on a zinc selenide (ZnSe) substrate. The effects of the ZnSe surface morphology (specifically, two variations of nanometer-scale corrugations obtained by mechanical polishing) on the initial ordering and reorientation dynamics of the electric-field-induced Freedericksz transition are presented here. The interaction of the 5CB with ZnSe surfaces bearing a spicular corrugation induces a homeotropic (surface normal) alignment of the film confined in the cavity. Alternately, when ZnSe is polished to generate fine grooves along the surface, a planar alignment is promoted in the liquid crystalline film. Time-resolved FTIR studies that enable the direct measurement of the rate constants for the electric-field-induced orientation and thermal relaxation reveal that the dynamic transitions of the two film structures are significantly different. These measurements quantitatively demonstrate the strong effects of surface morphology on the anchoring, order, and dynamics of liquid crystalline thin films
Cross-Task Representation Learning for Anatomical Landmark Detection
Recently, there is an increasing demand for automatically detecting
anatomical landmarks which provide rich structural information to facilitate
subsequent medical image analysis. Current methods related to this task often
leverage the power of deep neural networks, while a major challenge in fine
tuning such models in medical applications arises from insufficient number of
labeled samples. To address this, we propose to regularize the knowledge
transfer across source and target tasks through cross-task representation
learning. The proposed method is demonstrated for extracting facial anatomical
landmarks which facilitate the diagnosis of fetal alcohol syndrome. The source
and target tasks in this work are face recognition and landmark detection,
respectively. The main idea of the proposed method is to retain the feature
representations of the source model on the target task data, and to leverage
them as an additional source of supervisory signals for regularizing the target
model learning, thereby improving its performance under limited training
samples. Concretely, we present two approaches for the proposed representation
learning by constraining either final or intermediate model features on the
target model. Experimental results on a clinical face image dataset demonstrate
that the proposed approach works well with few labeled data, and outperforms
other compared approaches.Comment: MICCAI-MLMI 202
Intraoperative Organ Motion Models with an Ensemble of Conditional Generative Adversarial Networks
In this paper, we describe how a patient-specific, ultrasound-probe-induced
prostate motion model can be directly generated from a single preoperative MR
image. Our motion model allows for sampling from the conditional distribution
of dense displacement fields, is encoded by a generative neural network
conditioned on a medical image, and accepts random noise as additional input.
The generative network is trained by a minimax optimisation with a second
discriminative neural network, tasked to distinguish generated samples from
training motion data. In this work, we propose that 1) jointly optimising a
third conditioning neural network that pre-processes the input image, can
effectively extract patient-specific features for conditioning; and 2)
combining multiple generative models trained separately with heuristically
pre-disjointed training data sets can adequately mitigate the problem of mode
collapse. Trained with diagnostic T2-weighted MR images from 143 real patients
and 73,216 3D dense displacement fields from finite element simulations of
intraoperative prostate motion due to transrectal ultrasound probe pressure,
the proposed models produced physically-plausible patient-specific motion of
prostate glands. The ability to capture biomechanically simulated motion was
evaluated using two errors representing generalisability and specificity of the
model. The median values, calculated from a 10-fold cross-validation, were
2.8+/-0.3 mm and 1.7+/-0.1 mm, respectively. We conclude that the introduced
approach demonstrates the feasibility of applying state-of-the-art machine
learning algorithms to generate organ motion models from patient images, and
shows significant promise for future research.Comment: Accepted to MICCAI 201
Adversarial Deformation Regularization for Training Image Registration Neural Networks
We describe an adversarial learning approach to constrain convolutional
neural network training for image registration, replacing heuristic smoothness
measures of displacement fields often used in these tasks. Using
minimally-invasive prostate cancer intervention as an example application, we
demonstrate the feasibility of utilizing biomechanical simulations to
regularize a weakly-supervised anatomical-label-driven registration network for
aligning pre-procedural magnetic resonance (MR) and 3D intra-procedural
transrectal ultrasound (TRUS) images. A discriminator network is optimized to
distinguish the registration-predicted displacement fields from the motion data
simulated by finite element analysis. During training, the registration network
simultaneously aims to maximize similarity between anatomical labels that
drives image alignment and to minimize an adversarial generator loss that
measures divergence between the predicted- and simulated deformation. The
end-to-end trained network enables efficient and fully-automated registration
that only requires an MR and TRUS image pair as input, without anatomical
labels or simulated data during inference. 108 pairs of labelled MR and TRUS
images from 76 prostate cancer patients and 71,500 nonlinear finite-element
simulations from 143 different patients were used for this study. We show that,
with only gland segmentation as training labels, the proposed method can help
predict physically plausible deformation without any other smoothness penalty.
Based on cross-validation experiments using 834 pairs of independent validation
landmarks, the proposed adversarial-regularized registration achieved a target
registration error of 6.3 mm that is significantly lower than those from
several other regularization methods.Comment: Accepted to MICCAI 201
Self-supervised Representation Learning for Ultrasound Video
Recent advances in deep learning have achieved promising performance for
medical image analysis, while in most cases ground-truth annotations from human
experts are necessary to train the deep model. In practice, such annotations
are expensive to collect and can be scarce for medical imaging applications.
Therefore, there is significant interest in learning representations from
unlabelled raw data. In this paper, we propose a self-supervised learning
approach to learn meaningful and transferable representations from medical
imaging video without any type of human annotation. We assume that in order to
learn such a representation, the model should identify anatomical structures
from the unlabelled data. Therefore we force the model to address anatomy-aware
tasks with free supervision from the data itself. Specifically, the model is
designed to correct the order of a reshuffled video clip and at the same time
predict the geometric transformation applied to the video clip. Experiments on
fetal ultrasound video show that the proposed approach can effectively learn
meaningful and strong representations, which transfer well to downstream tasks
like standard plane detection and saliency prediction.Comment: ISBI 202
Self-Supervised Ultrasound to MRI Fetal Brain Image Synthesis
Fetal brain magnetic resonance imaging (MRI) offers exquisite images of the
developing brain but is not suitable for second-trimester anomaly screening,
for which ultrasound (US) is employed. Although expert sonographers are adept
at reading US images, MR images which closely resemble anatomical images are
much easier for non-experts to interpret. Thus in this paper we propose to
generate MR-like images directly from clinical US images. In medical image
analysis such a capability is potentially useful as well, for instance for
automatic US-MRI registration and fusion. The proposed model is end-to-end
trainable and self-supervised without any external annotations. Specifically,
based on an assumption that the US and MRI data share a similar anatomical
latent space, we first utilise a network to extract the shared latent features,
which are then used for MRI synthesis. Since paired data is unavailable for our
study (and rare in practice), pixel-level constraints are infeasible to apply.
We instead propose to enforce the distributions to be statistically
indistinguishable, by adversarial learning in both the image domain and feature
space. To regularise the anatomical structures between US and MRI during
synthesis, we further propose an adversarial structural constraint. A new
cross-modal attention technique is proposed to utilise non-local spatial
information, by encouraging multi-modal knowledge fusion and propagation. We
extend the approach to consider the case where 3D auxiliary information (e.g.,
3D neighbours and a 3D location index) from volumetric data is also available,
and show that this improves image synthesis. The proposed approach is evaluated
quantitatively and qualitatively with comparison to real fetal MR images and
other approaches to synthesis, demonstrating its feasibility of synthesising
realistic MR images.Comment: IEEE Transactions on Medical Imaging 202
- …